Using Pre-Trained Language Models for Abstractive DBPEDIA Summarization: A Comparative Study

نویسندگان

چکیده

Purpose: This study addresses the limitations of current short abstracts DBPEDIA entities, which often lack a comprehensive overview due to their creating method (i.e., selecting first two-three sentences from full abstracts). Methodology: We leverage pre-trained language models generate abstractive summaries in six languages (English, French, German, Italian, Spanish, and Dutch). performed several experiments assess quality generated by models. In particular, we evaluated using human judgments automated metrics (Self-ROUGE BERTScore). Additionally, studied correlation between evaluating under different aspects: informativeness, coherence, conciseness, fluency. Findings: Pre-trained more concise informative than existing abstracts. Specifically, BART-based effectively overcome abstracts, especially for longer ones. Moreover, show that BERTScore ROUGE-1 are reliable assessing informativeness coherence with respect also find negative conciseness ratings. Furthermore, fluency evaluation remains challenging without judgment. Value: has significant implications various applications machine learning natural processing rely on resources. By providing succinct summaries, our approach enhances contributes semantic web community.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Multilingual Natural Language Generation within Abstractive Summarization

With the tremendous amount of textual data available in the Internet, techniques for abstractive text summarization become increasingly appreciated. In this paper, we present work in progress that tackles the problem of multilingual text summarization using semantic representations. Our system is based on abstract linguistic structures obtained from an analysis pipeline of disambiguation, synta...

متن کامل

TL;DR: Improving Abstractive Summarization Using LSTMs

Traditionally, summarization has been approached through extractive methods. However, they have produced limited results. More recently, neural sequence-tosequence models for abstractive text summarization have shown more promise, although the task still proves to be challenging. In this paper, we explore current state-of-the-art architectures and reimplement them from scratch. We begin with a ...

متن کامل

Text Generation for Abstractive Summarization

We have begun work on a framework for abstractive summarization and decided to focus on a module for text generation. For TAC 2010, we thus move away from sentence extraction. Each sentence in the summary we generate is based on a document sentence but it usually contains a smaller amount of information and uses fewer words. The system uses the output of a syntactic parser for a sentence and th...

متن کامل

Abstractive Meeting Summarization Using Dependency Graph Fusion

ive Meeting Summarization Using Dependency Graph Fusion Siddhartha Banerjee The Pennsylvania State University University Park PA, USA 16802 [email protected] Prasenjit Mitra Qatar Computing Research Institute Tornado Tower, 18th floor Doha, Qatar [email protected] Kazunari Sugiyama National University of Singapore 13 Computing Drive Singapore 117417 [email protected]

متن کامل

A Deep Reinforced Model for Abstractive Summarization

Attentional, RNN-based encoder-decoder models for abstractive summarization have achieved good performance on short input and output sequences. For longer documents and summaries however these models often include repetitive and incoherent phrases. We introduce a neural network model with a novel intraattention that attends over the input and continuously generated output separately, and a new ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Studies on the semantic web

سال: 2023

ISSN: ['1868-1158']

DOI: https://doi.org/10.3233/ssw230003